Open Problem: The Oracle Complexity of Smooth Convex Optimization in Nonstandard Settings

نویسنده

  • Cristobal Guzman
چکیده

First-order convex minimization algorithms are currently the methods of choice for large-scale sparse – and more generally parsimonious – regression models. We pose the question on the limits of performance of black-box oriented methods for convex minimization in non-standard settings, where the regularity of the objective is measured in a norm not necessarily induced by the feasible domain. This question is studied for `p/`q-settings, and their matrix analogues (Schatten norms), where we find surprising gaps on lower bounds compared to state of the art methods. We propose a conjecture on the optimal convergence rates for these settings, for which a positive answer would lead to significant improvements on minimization algorithms for parsimonious regression models. Let (E, ‖ · ‖) be a finite-dimensional normed space. Given parameters, 1 < κ ≤ 2, and L > 0, we consider the class F‖·‖(κ, L) of convex functions that are (κ, L)-smooth w.r.t. norm ‖ · ‖. One such function f : E→ R satisfies that ‖∇f(y)−∇f(x)‖∗ ≤ L‖x− y‖κ−1 ∀x, y ∈ E. Notice that the case κ → 1 corresponds essentially to nonsmooth (Lipschitz continuous) convex functions, and κ = 2 corresponds to smooth (with Lipschitz continuous gradients) convex functions. Given a convex body X ⊆ E, we are interested on the complexity of the problem class P = (F‖·‖(κ, L), X), comprised of optimization problems with objective f ∈ F‖·‖(κ, L) Opt(f,X) = min{f(x) : x ∈ X}. (Pf,X) We study a black-box oracle model where most algorithms based on subgradient computations can be implemented1. Here, an algorithm is allowed to perform queries x ∈ E, and for any such query the oracle returns Of (x) (e.g., for first-order methods, Of (x) = ∇f(x)). The only assumption on the oracle is locality: For any x ∈ E and f, g ∈ F such that there exists a neighborhood B(x, δ) where f ≡ g, then Of (x) = Og(x). Given T > 0, we consider an algorithm A whose output xT (A, f) is only determined by T (adaptive) oracle queries. We define the accuracy of algorithm A on an instance f as ε(A, f) := f(xT (A, f)) − Opt(f,X), if xT (A, f) ∈ X , otherwise ε(A, f) = ∞. We characterize optimal 1. Notable exceptions are methods exploiting explicit saddle-point description, e.g., the smoothing technique by Nesterov (2005). Note however that such algorithms do not give improvement in the smooth case.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Tight Complexity Bounds for Optimizing Composite Objectives

We provide tight upper and lower bounds on the complexity of minimizing the average of m convex functions using gradient and prox oracles of the component functions. We show a significant gap between the complexity of deterministic vs randomized optimization. For smooth functions, we show that accelerated gradient descent (AGD) and an accelerated variant of SVRG are optimal in the deterministic...

متن کامل

Oracle Complexity of Second-Order Methods for Smooth Convex Optimization

Second-order methods, which utilize gradients as well as Hessians to optimize a given function, are of major importance in mathematical optimization. In this work, we study the oracle complexity of such methods, or equivalently, the number of iterations required to optimize a function to a given accuracy. Focusing on smooth and convex functions, we derive (to the best of our knowledge) the firs...

متن کامل

A Linearly Convergent Conditional Gradient Algorithm with Applications to Online and Stochastic Optimization

Linear optimization is many times algorithmically simpler than non-linear convex optimization. Linear optimization over matroid polytopes, matching polytopes and path polytopes are example of problems for which we have simple and efficient combinatorial algorithms, but whose non-linear convex counterpart is harder and admit significantly less efficient algorithms. This motivates the computation...

متن کامل

First-order methods of smooth convex optimization with inexact oracle

In this paper, we analyze different first-order methods of smooth convex optimization employing inexact first-order information. We introduce the notion of an approximate first-order oracle. The list of examples of such an oracle includes smoothing technique, Moreau-Yosida regularization, Modified Lagrangians, and many others. For different methods, we derive complexity estimates and study the ...

متن کامل

حل مسئله پخش بار بهینه در شرایط نرمال و اضطراری با استفاده از الگوریتم ترکیبی گروه ذرات و نلدر مید (PSO-NM)

In this paper, solving optimal power flow problem has been investigated by using hybrid particle swarm optimization and Nelder Mead Algorithms. The goal of combining Nelder-Mead (NM) simplex method and particle swarm optimization (PSO) is to integrate their advantages and avoid their disadvantages. NM simplex method is a very efficient local search procedure but its convergence is extremely sen...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2015